Multi-Task and Transfer Learning with Recurrent Neural Networks
نویسنده
چکیده
Being able to obtain a model of a dynamical system is useful for a variety of purposes, e.g. condition monitoring and model-based control. The dynamics of complex technical systems such as gas or wind turbines can be approximated by data driven models, e.g. recurrent neural networks. Such methods have proven to be powerful alternatives to analytical models which are not always available or may be inaccurate. Optimizing the parameters of data-driven models generally requires large amounts of operational data. However, data is a scarce resource in many applications, hence, data-efficient procedures utilizing all available data are preferred. In this thesis, recurrent neural networks are explored and developed which allow for data-efficient knowledge transfer from one or multiple source tasks to a related target task that lacks sufficient amounts of data. Multiple knowledge transfer scenarios are considered: dual-task learning, multi-task learning, and transfer learning. Dual-task learning is a particular case of multi-task learning in which multiple tasks are learned simultaneously, thus, allowing to share knowledge among the tasks. Transfer learning implies a sequential protocol in which the target task is learned subsequent to learning one or multiple source tasks. Knowledge transfer is explored and applied in the context of fully and partially observable system identification to improve the model of a little observed system by means of auxiliary data from one or multiple similar systems. Fully observable system identification assumes that the observed system state corresponds to the true state while partial observability implies the observation of either a subset of the state variables or proxy variables. The primary contribution of this thesis comprises a novel recurrent neural network architecture for knowledge transfer which uses factored third-order tensors to encode cross-task and task-specific information within composite affine transformations. This model is investigated in a series of experiments on synthetic
منابع مشابه
Multi-Step-Ahead Prediction of Stock Price Using a New Architecture of Neural Networks
Modelling and forecasting Stock market is a challenging task for economists and engineers since it has a dynamic structure and nonlinear characteristic. This nonlinearity affects the efficiency of the price characteristics. Using an Artificial Neural Network (ANN) is a proper way to model this nonlinearity and it has been used successfully in one-step-ahead and multi-step-ahead prediction of di...
متن کاملData-Efficient Temporal Regression with Multi-Task Recurrent Neural Networks
We demonstrate the utility of multi-task learning with recurrent neural networks in the context of a temporal regression task using real world gas turbine data. Our goal is to learn the input-output relationship of a task with temporal dependencies given only few data by exploiting additional data of related tasks. Therefore, we propose a multi-task learning approach with many inter-task and fe...
متن کاملRepresentation Stability as a Regularizer for Improved Text Analytics Transfer Learning
Although neural networks are well suited for sequential transfer learning tasks, the catastrophic forgetting problem hinders proper integration of prior knowledge. In this work, we propose a solution to this problem by using a multi-task objective based on the idea of distillation and a mechanism that directly penalizes forgetting at the shared representation layer during the knowledge integrat...
متن کاملExploring Cross-Lingual Transfer of Morphological Knowledge In Sequence-to-Sequence Models
Multi-task training is an effective method to mitigate the data sparsity problem. It has recently been applied for crosslingual transfer learning for paradigm completion—the task of producing inflected forms of lemmata—with sequenceto-sequence networks. However, it is still vague how the model transfers knowledge across languages, as well as if and which information is shared. To investigate th...
متن کاملMulti-task learning strategies for a recurrent neural net in a hybrid tied-posteriors acoustic model
An important goal of an automatic classifier is to learn the best possible generalization from given training material. One possible improvement over a standard learning algorithm is to train several related tasks in parallel. We apply the multi-task learning scheme to a recurrent neural network estimating phoneme posterior probabilities and HMM state posterior probabilities, respectively. A co...
متن کامل